Stochastic Block BFGS: Squeezing More Curvature out of Data

نویسندگان

  • Robert M. Gower
  • Donald Goldfarb
  • Peter Richtárik
چکیده

is cheap (2), where Dt ∈ Rd×q and q min{d, n}. We employ three di erent sketching strategies: 1) gauss. Dt has standard Gaussian entries sampled i.i.d at each iteration. 2) prev. Let dt = −Htgt. Store search directions Dt = [dt+1−q , . . . , dt] and update Ht once every q iterations. 3) fact. Sample Ct ⊆ {1, . . . , d} uniformly at random and set Dt = Lt−1I:Ct ,where Lt−1L T t−1 = Ht−1 and I:Ct denotes the concatenation of the columns of the identity matrix indexed by a set Ct ⊂ {1, . . . , d}. 4. Block BFGS Update

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The modified BFGS method with new secant relation ‎for unconstrained optimization problems‎

Using Taylor's series we propose a modified secant relation to get a more accurate approximation of the second curvature of the objective function. Then, based on this modified secant relation we present a new BFGS method for solving unconstrained optimization problems. The proposed method make use of both gradient and function values while the usual secant relation uses only gradient values. U...

متن کامل

Global convergence of online limited memory BFGS

Global convergence of an online (stochastic) limited memory version of the Broyden-FletcherGoldfarb-Shanno (BFGS) quasi-Newton method for solving optimization problems with stochastic objectives that arise in large scale machine learning is established. Lower and upper bounds on the Hessian eigenvalues of the sample functions are shown to suffice to guarantee that the curvature approximation ma...

متن کامل

On the Use of Stochastic Hessian Information in Optimization Methods for Machine Learning

This paper describes how to incorporate sampled curvature information in a NewtonCG method and in a limited memory quasi-Newton method for statistical learning. The motivation for this work stems from supervised machine learning applications involving a very large number of training points. We follow a batch approach, also known in the stochastic optimization literature as a sample average appr...

متن کامل

Block Bfgs Methods

We introduce a quasi-Newton method with block updates called Block BFGS. We show that this method, performed with inexact Armijo-Wolfe line searches, converges globally and superlinearly under the same convexity assumptions as BFGS. We also show that Block BFGS is globally convergent to a stationary point when applied to non-convex functions with bounded Hessian, and discuss other modifications...

متن کامل

A Stochastic Quasi-Newton Method for Non-Rigid Image Registration

Image registration is often very slow because of the high dimensionality of the images and complexity of the algorithms. Adaptive stochastic gradient descent (ASGD) outperforms deterministic gradient descent and even quasi-Newton in terms of speed. This method, however, only exploits first-order information of the cost function. In this paper, we explore a stochastic quasi-Newton method (s-LBFG...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016